Chen Wang
by Chen Wang

Categories

  • docs
  • Machine Learning

Tags

  • machine learning
  • transparency
  • interaction

1. Introduce the Main Idea of Transparency and Interaction in Machine Learning

Transparency

Transparency aims to understand what did the model learn, what do the weights tell us about features and their importance, when an object is classified by this model, which feature values contributed to each class and how much.

Interaction

If we want to label more objects, which one should we label next? Can we provide any rationales into the learning process?

Why Transparency?

A perfect example comes from my professor Mustafa. He said that sometimes the machine instead provides a suggestion and gives a reason than deciding for the human. In medical recommender system, the doctor prefers the system provides suspected symptoms to the percentage of the diagnosis. If a cancer detection system only gives “This patient has a 79% chance gets cancer”. Such a result will make confuse the doctor and the patient. In contrast, if this system said: “This patient has a 79% chance gets cancer because I detected the patient has high compression in the spinal cord and pleural effusion”. This diagonal report will be of great help to doctors and dramatically reduce diagnostic time. Moreover, transparency also used in Trustworthy fraud detection{target:blank}. We should tell the customer this product evaluation is fake and why at the same time.

2. Concept Learning

Motivation

Different classification task has different purposes such as emails labeled as spam/~spam, patient records labeled as flu/~flu or reviews labeled as positive/negative.

Our goal is inducing a general function that fits the training data well and generalized well to unseen or future data.

General-to-Specific Ordering of Hypothese

Example

Suppose we have two features: Weight: (Light, Heavy) and Color: (Red, Green, Blue. The class is YES and NO. Therefore, we have six objects/instances . and we have 64 hypotheses/functions .

Weight Color F1 F64
Light Red NO   YES
Light Green NO   YES
Light Blue NO   YES
Heavy Red NO   YES
Heavy Green NO   YES
Heavy Blue NO   YES

If we know that <Light, Red> is ‘Yes’, can you guess how many hypothese left? The answer is .

The number of hypotheses of <Light, Blue > is ‘Yes,’ or ‘No’ is 16.

For now, we know how to calculate the hypothesis. Let us assume two particular assumptions:

  1. ‘No’ to everything
  2. ‘Yes’ to everything
  • In particular:
    • : ‘No’ to everything
    • : ‘Yes’ to everything
    • <Light, Red>: ‘Yes’ to <Light, Red> and ‘No’ to everthing else
    • <Light, >: ‘Yes’ to anything that is Light; ignore Color; everything else is ‘No’
    • Combination of one or more of with other feature values is still equivalen to ‘No’ to everything. E.g., Red is ‘No’ to everything
  • Most-general hypothesis is where :
  • Most-specific hypothesis is where :
  • These called hypothesis space

But, how do we search this space efficiently?

Algorithms

1.Find-S

Initialize to the most specific hypothesis in For each positive training instance

  • For each attribute constraint in :
    • If the constraint in is satisfied by
    • Then do nothing
    • Else replace in by the next more general constraint that is satisfied by Output hypothesis
  • Warning!!!: In that case, Find-S always return the most specific one, but we do not know whether we need the most specific one or the most general one or interval between them. Moreover, what if our training hata has errors or what if the most specific hypothesis in not unique?

2.List-Then-Eliminate

VS a list of containing every hypothesis in For each training example,

  • Remove from VS any hypothesis for which Output VS

Though this algorithm can output all consistent hypotheses, we need to list all possible explanation. It is possible in small predictions space. However, it is impossible for infinite hypothesis spaces.

3.Candidate-Elimination

Initialize to the set of maximally-general hypotheses in Initialize to the set of maximally-specific hypotheses in For each training example , do

  • If is a positive example
    • Remove from any hypothesis that is inconsistent with
    • For each hypothesis in that is not consistent with
      • Remove from
      • Add to to all minimal gegneralizations of such that
        • is consistent with , and some member of is more general than
      • Remove from any hypothesis that is more general than another hypothesis in
Version Space, Source: Wikipedia